SVM Regression Modelling and its Applications

نویسندگان

  • Jos De Brabanter
  • Joos Vandewalle
چکیده

The key method in this thesis is least squares support vector machines (LSSVM), a class of kernel based learning methods that fits within the penalized modelling paradigm. Primary goals of the LS-SVM models are regression and classification. Although local methods (kernel methods) focus directly on estimating the function at a point, they face problems in high dimensions. Therefore, one can guarantee good estimation of a high-dimensional function only if the function is extremely smooth. We have incorporated additional assumptions (the regression function is an additive function of its components) to overcome the curse of dimensionality. We have studied the properties of the LS-SVM regression when relaxing the Gauss-Markov conditions. It was recognized that outliers may have an unusually large influence on the resulting estimate. However, asymptotically the heteroscedasticity does not play any important role. We have developed a robust framework for LS-SVM regression. It allows to obtain a robust estimate based upon the previous LS-SVM regression solution, in a subsequent step. The weights are determined based upon the distribution of the error variables. We have shown, based on the empirical influence curve and the maxbias curve, that the weighted LS-SVM regression is a robust function estimation tool. We have used the same principle to obtain an LS-SVM regression estimate in the heteroscedastic case. However, the weights are then based upon a smooth error variance estimate. Most efficient learning algorithms in neural networks, support vector machines and kernel based learning methods require the tuning of some extra tuning parameters. For practical use, it is often preferable to have a data-driven method to select these parameters. Based on location estimators (e.g., mean, median, M-estimators, L-estimators, R-estimators), we have introduced robust counterparts of model selection criteria (e.g., Cross-Validation, Final Prediction Error criterion). Inference procedures for both linear and nonlinear parametric regression models in fact assume that the output variable follows a normal distribution. With nonparametric regression, the regression equation is determined from the data. In this case, we relax the normality assumption and standard inference procedures are no longer applicable in that case. We have developed a robust approach for obtaining robust prediction intervals by using robust external bootstrapping methods. Finally, we apply LS-SVM regression modelling in the case of density estimation.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Support Vector Machines: Review and Applications in Civil Engineering

Support Vector Machines (SVM) are emerging techniques which have proven successful in many traditionally neural network (NN) dominated applications. An interesting property of this approach is that it is an approximate implementation of the Structural Risk Minimization (SRM) induction principle that aims at minimizing a bound on the generalization error of a model, rather than minimizing the me...

متن کامل

A New Five-Parameter Distribution: Properties and Applications

In this paper, a new five-parameter lifetime and reliability distribution named “the exponentiated Uniform-Pareto distribution (EU-PD),” has been suggested that it has a bathtub-shaped and inverse bathtub-shape for modeling lifetime data. This distribution has applications in economics, actuarial modelling, reliability modeling, lifetime and biological sciences. Firstly, the mathematical and st...

متن کامل

Modified Logistic Regression: An Approximation to SVM and Its Applications in Large-Scale Text Categorization

Logistic Regression (LR) has been widely used in statistics for many years, and has received extensive study in machine learning community recently due to its close relations to Support Vector Machines (SVM) and AdaBoost. In this paper, we use a modified version of LR to approximate the optimization of SVM by a sequence of unconstrained optimization problems. We prove that our approximation wil...

متن کامل

Application an Svm Machine to Financial Time Series Modelling

Abstract: In Support Vector Machines (SVM ́s), a non-linear model is estimated based on solving a Quadratic Programming (QP) problem. Based on work [1] we investigate the quantifying of statistical structural model parameters of inflation in Slovak economics. Dynamic and SVM ́s modelling approaches are used for automated specification of a functional form of the model in data mining systems. Base...

متن کامل

Efficient Optimization of the Parameters of LS-SVM for Regression versus Cross-Validation Error

Least Squares Support Vector Machines (LS-SVM) are the state of the art in kernel methods for regression and function approximation. In the last few years, these models have been successfully applied to time series modelling and prediction. A key issue for the good performance of a LS-SVM model are the values chosen for both the kernel parameters and its hyperparameters in order to avoid overfi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2004